- Home
- Search Results
- Page 1 of 1
Search for: All records
-
Total Resources4
- Resource Type
-
0002000002000000
- More
- Availability
-
40
- Author / Contributor
- Filter by Author / Creator
-
-
Liang, Tianyu (4)
-
Biros, George (2)
-
Chen, Chao (2)
-
Anslyn, Eric V. (1)
-
Buluc, Aydin (1)
-
Demmel, James (1)
-
Du, Meiqing (1)
-
Hu, Wei (1)
-
Martinsson, Per-Gunnar (1)
-
Murray, Riley (1)
-
Sun, Xiaolong (1)
-
Wu, Tianhong (1)
-
Zhang, Sijia (1)
-
Zhang, Yanfeng (1)
-
#Tyler Phillips, Kenneth E. (0)
-
#Willis, Ciara (0)
-
& Abreu-Ramos, E. D. (0)
-
& Abramson, C. I. (0)
-
& Abreu-Ramos, E. D. (0)
-
& Adams, S.G. (0)
-
- Filter by Editor
-
-
& Spizer, S. M. (0)
-
& . Spizer, S. (0)
-
& Ahn, J. (0)
-
& Bateiha, S. (0)
-
& Bosch, N. (0)
-
& Brennan K. (0)
-
& Brennan, K. (0)
-
& Chen, B. (0)
-
& Chen, Bodong (0)
-
& Drown, S. (0)
-
& Ferretti, F. (0)
-
& Higgins, A. (0)
-
& J. Peters (0)
-
& Kali, Y. (0)
-
& Ruiz-Arias, P.M. (0)
-
& S. Spitzer (0)
-
& Sahin. I. (0)
-
& Spitzer, S. (0)
-
& Spitzer, S.M. (0)
-
(submitted - in Review for IEEE ICASSP-2024) (0)
-
-
Have feedback or suggestions for a way to improve these results?
!
Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
This work focuses on accelerating the multiplication of a dense random matrix with a (fixed) sparse matrix, which is frequently used in sketching algorithms. We develop a novel scheme that takes advantage of blocking and recomputation (on- the-fly random number generation) to accelerate this operation. The techniques we propose decrease memory movement, thereby increasing the algorithm’s parallel scalability in shared memory architectures. On the Intel Frontera architecture, our algorithm can achieve 2x speedups over libraries such as Eigen and Intel MKL on some examples. In addition, with 32 threads, we can obtain a parallel efficiency of up to approximately 45%. We also present a theoretical analysis for the memory movement lower bound of our algorithm, showing that under mild assumptions, it's possible to beat the data movement lower bound of general matrix-matrix multiply (GEMM) by a factor of sqrt(M), where $$M$$ is the cache size. Finally, we incorporate our sketching method into a randomized algorithm for overdetermined least squares with sparse data matrices. Our results are competitive with SuiteSparse for highly overdetermined problems; in some cases, we obtain a speedup of 10x over SuiteSparse.more » « less
-
Liang, Tianyu; Chen, Chao; Martinsson, Per-Gunnar; Biros, George (, IEEE)
-
Chen, Chao; Liang, Tianyu; Biros, George (, SIAM Journal on Scientific Computing)
-
Wu, Tianhong; Liang, Tianyu; Hu, Wei; Du, Meiqing; Zhang, Sijia; Zhang, Yanfeng; Anslyn, Eric V.; Sun, Xiaolong (, ACS Macro Letters)
An official website of the United States government
